3,294 research outputs found
Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors
Although the quest for more accurate solutions is pushing deep learning
research towards larger and more complex algorithms, edge devices demand
efficient inference and therefore reduction in model size, latency and energy
consumption. One technique to limit model size is quantization, which implies
using fewer bits to represent weights and biases. Such an approach usually
results in a decline in performance. Here, we introduce a method for designing
optimally heterogeneously quantized versions of deep neural network models for
minimum-energy, high-accuracy, nanosecond inference and fully automated
deployment on chip. With a per-layer, per-parameter type automatic quantization
procedure, sampling from a wide range of quantizers, model energy consumption
and size are minimized while high accuracy is maintained. This is crucial for
the event selection procedure in proton-proton collisions at the CERN Large
Hadron Collider, where resources are strictly limited and a latency of
s is required. Nanosecond inference and a resource
consumption reduced by a factor of 50 when implemented on field-programmable
gate array hardware are achieved
Fast convolutional neural networks on FPGAs with hls4ml
We introduce an automated tool for deploying ultra low-latency, low-power
deep neural networks with convolutional layers on FPGAs. By extending the
hls4ml library, we demonstrate an inference latency of s using
convolutional architectures, targeting microsecond latency applications like
those at the CERN Large Hadron Collider. Considering benchmark models trained
on the Street View House Numbers Dataset, we demonstrate various methods for
model compression in order to fit the computational constraints of a typical
FPGA device used in trigger and data acquisition systems of particle detectors.
In particular, we discuss pruning and quantization-aware training, and
demonstrate how resource utilization can be significantly reduced with little
to no loss in model accuracy. We show that the FPGA critical resource
consumption can be reduced by 97% with zero loss in model accuracy, and by 99%
when tolerating a 6% accuracy degradation.Comment: 18 pages, 18 figures, 4 table
Combination of searches for heavy resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collisions at = 8 and 13 TeV
A statistical combination of searches is presented for massive resonances decaying to WW, WZ, ZZ, WH, and ZH boson pairs in proton–proton collision data collected by the CMS experiment at the LHC. The data were taken at centre-of-mass energies of 8 and 13 TeV, corresponding to respective integrated luminosities of 19.7 and up to 2.7 . The results are interpreted in the context of heavy vector triplet and singlet models that mimic properties of composite-Higgs models predicting W′ and Z′ bosons decaying to WZ, WW, WH, and ZH bosons. A model with a bulk graviton that decays into WW and ZZ is also considered. This is the first combined search for WW, WZ, WH, and ZH resonances and yields lower limits on masses at 95% confidence level for W′ and Z′ singlets at 2.3 TeV, and for a triplet at 2.4 TeV. The limits on the production cross section of a narrow bulk graviton resonance with the curvature scale of the warped extra dimension , in the mass range of 0.6 to 4.0 TeV, are the most stringent published to date
Search for supersymmetry in pp collisions at = 13 tev in the single-lepton final state using the sum of masses of large-radius jets
Results are reported from a search for supersymmetric particles in proton-proton collisions in the final state with a single lepton, multiple jets, including at least one b-tagged jet, and large missing transverse momentum. The search uses a sample of proton-proton collision data at = 13 TeV recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 35.9  . The observed event yields in the signal regions are consistent with those expected from standard model backgrounds. The results are interpreted in the context of simplified models of supersymmetry involving gluino pair production, with gluino decay into either on- or off-mass-shell top squarks. Assuming that the top squarks decay into a top quark plus a stable, weakly interacting neutralino, scenarios with gluino masses up to about 1.9 TeV are excluded at 95% confidence level for neutralino masses up to about 1 TeV
Search for dark matter produced in association with heavy-flavor quark pairs in proton-proton collisions at = 13 TeV
A search is presented for an excess of events with heavy-flavor quark pairs ( and ) and a large imbalance in transverse momentum in data from proton–proton collisions at a center-of-mass energy of 13TeV. The data correspond to an integrated luminosity of 2.2 collected with the CMS detector at the CERN LHC. No deviations are observed with respect to standard model predictions. The results are used in the first interpretation of dark matter production in and final states in a simplified model. This analysis is also the first to perform a statistical combination of searches for dark matter produced with different heavy-flavor final states. The combination provides exclusions that are stronger than those achieved with individual heavy-flavor final states
Accelerated Charged Particle Tracking with Graph Neural Networks on FPGAs
We develop and study FPGA implementations of algorithms for charged particle
tracking based on graph neural networks. The two complementary FPGA designs are
based on OpenCL, a framework for writing programs that execute across
heterogeneous platforms, and hls4ml, a high-level-synthesis-based compiler for
neural network to firmware conversion. We evaluate and compare the resource
usage, latency, and tracking performance of our implementations based on a
benchmark dataset. We find a considerable speedup over CPU-based execution is
possible, potentially enabling such algorithms to be used effectively in future
computing workflows and the FPGA-based Level-1 trigger at the CERN Large Hadron
Collider.Comment: 8 pages, 4 figures, To appear in Third Workshop on Machine Learning
and the Physical Sciences (NeurIPS 2020
The CMS trigger system
This paper describes the CMS trigger system and its performance during Run 1 of the LHC. The trigger system consists of two levels designed to select events of potential physics interest from a GHz (MHz) interaction rate of proton-proton (heavy ion) collisions. The first level of the trigger is implemented in hardware, and selects events containing detector signals consistent with an electron, photon, muon, Ï„ lepton, jet, or missing transverse energy. A programmable menu of up to 128 object-based algorithms is used to select events for subsequent processing. The trigger thresholds are adjusted to the LHC instantaneous luminosity during data taking in order to restrict the output rate to 100 kHz, the upper limit imposed by the CMS readout electronics. The second level, implemented in software, further refines the purity of the output stream, selecting an average rate of 400 Hz for offline event storage. The objectives, strategy and performance of the trigger system during the LHC Run 1 are described
Search for dark matter produced with an energetic jet or a hadronically decaying W or Z boson at TeV
A search for dark matter particles is performed using events with large missing transverse momentum, at least one energetic jet, and no leptons, in proton-proton collisions at TeV collected with the CMS detector at the LHC. The data sample corresponds to an integrated luminosity of 12.9 . The search includes events with jets from the hadronic decays of a W or Z boson. The data are found to be in agreement with the predicted background contributions from standard model processes. The results are presented in terms of simplified models in which dark matter particles are produced through interactions involving a vector, axial-vector, scalar, or pseudoscalar mediator. Vector and axial-vector mediator particles with masses up to 1.95 TeV, and scalar and pseudoscalar mediator particles with masses up to 100 and 430 GeV respectively, are excluded at 95% confidence level. The results are also interpreted in terms of the invisible decays of the Higgs boson, yielding an observed (expected) 95% confidence level upper limit of 0.44 (0.56) on the corresponding branching fraction. The results of this search provide the strongest constraints on the dark matter pair production cross section through vector and axial-vector mediators at a particle collider. When compared to the direct detection experiments, the limits obtained from this search provide stronger constraints for dark matter masses less than 5, 9, and 550 GeV, assuming vector, scalar, and axial-vector mediators, respectively. The search yields stronger constraints for dark matter masses less than 200 GeV, assuming a pseudoscalar mediator, when compared to the indirect detection results from Fermi-LAT
Search for a light pseudoscalar Higgs boson produced in association with bottom quarks in pp collisions at = 8 TeV
A search for a light pseudoscalar Higgs boson (A) produced in association with bottom quarks and decaying into a muon pair is reported. The search uses 19.7 of proton-proton collisions at a center-of-mass energy of 8 TeV, collected by the CMS experiment. No signal is observed in the dimuon mass range from 25 to 60 GeV. Upper limits on the cross section times branching fraction, , are set
- …